Context Memory Networks for Multi-objective Semantic Parsing in Conversational Understanding
نویسندگان
چکیده
The end-to-end multi-domain and multi-task learning of the full semantic frame of user utterances (i.e., domain and intent classes and slots in utterances) have recently emerged as a new paradigm in spoken language understanding. An advantage of the joint optimization of these semantic frames is that the data and feature representations learnt by the model are shared across different tasks (e.g., domain/intent classification and slot filling tasks use the same feature sets). It’s important that the model should learn to pay attention to global and local aspects of the utterances while learning to map the entire utterance to an intent class and tag each word with a slot tag. We introduce the Context Memory Network (CMN), a neural network architecture which specifically focuses on learning better representations as attention vectors from past memory to be reasoned with for the end task of jointly learning the intent class and slot tags. The utterances trigger a dynamic memory network, which learns attention based representation for each word by allowing the model to condition on the list of related phrases in the form of memory networks. These representations are then provided to a new multi-objective long short term memory network (LSTM) to infer the intent class and slot tags. Our empirical investigations on CMN show impressive gains over the end-to-end LSTM baselines on ATIS dataset as well as two other humanto-machine conversational datasets.
منابع مشابه
Improving Frame Semantic Parsing with Hierarchical Dialogue Encoders
Conversational Language Understanding (CLU) is a key component of goal oriented dialogue systems that would parse user utterances into semantic frame representations. Traditionally CLU does not utilize the dialogue history beyond the previous system turn and contextual ambiguities are resolved by the downstream components. In this paper, we explore novel approaches for modeling dialogue context...
متن کاملParsing Natural Language Conversations using Contextual Cues
In this work, we focus on semantic parsing of natural language conversations. Most existing methods for semantic parsing are based on understanding the semantics of a single sentence at a time. However, understanding conversations also requires an understanding of conversational context and discourse structure across sentences. We formulate semantic parsing of conversations as a structured pred...
متن کاملProgressively Diffused Networks for Semantic Image Segmentation
This paper introduces Progressively Diffused Networks (PDNs) for unifying multi-scale context modeling with deep feature learning, by taking semantic image segmentation as an exemplar application. Prior neural networks such as ResNet [11] tend to enhance representational power by increasing the depth of architectures and driving the training objective across layers. However, we argue that spati...
متن کاملTowards Zero-Shot Frame Semantic Parsing for Domain Scaling
State-of-the-art slot filling models for goal-oriented human/machine conversational language understanding systems rely on deep learning methods. While multi-task training of such models alleviates the need for large in-domain annotated datasets, bootstrapping a semantic parsing model for a new domain using only the semantic frame, such as the back-end API or knowledge graph schema, is still on...
متن کاملMulti-Domain Joint Semantic Frame Parsing Using Bi-Directional RNN-LSTM
Sequence-to-sequence deep learning has recently emerged as a new paradigm in supervised learning for spoken language understanding. However, most of the previous studies explored this framework for building single domain models for each task, such as slot filling or domain classification, comparing deep learning based approaches with conventional ones like conditional random fields. This paper ...
متن کامل